25 research outputs found

    Self-supervised Likelihood Estimation with Energy Guidance for Anomaly Segmentation in Urban Scenes

    Full text link
    Robust autonomous driving requires agents to accurately identify unexpected areas in urban scenes. To this end, some critical issues remain open: how to design advisable metric to measure anomalies, and how to properly generate training samples of anomaly data? Previous effort usually resorts to uncertainty estimation and sample synthesis from classification tasks, which ignore the context information and sometimes requires auxiliary datasets with fine-grained annotations. On the contrary, in this paper, we exploit the strong context-dependent nature of segmentation task and design an energy-guided self-supervised frameworks for anomaly segmentation, which optimizes an anomaly head by maximizing the likelihood of self-generated anomaly pixels. To this end, we design two estimators for anomaly likelihood estimation, one is a simple task-agnostic binary estimator and the other depicts anomaly likelihood as residual of task-oriented energy model. Based on proposed estimators, we further incorporate our framework with likelihood-guided mask refinement process to extract informative anomaly pixels for model training. We conduct extensive experiments on challenging Fishyscapes and Road Anomaly benchmarks, demonstrating that without any auxiliary data or synthetic models, our method can still achieves competitive performance to other SOTA schemes

    Learning with Noisy labels via Self-supervised Adversarial Noisy Masking

    Full text link
    Collecting large-scale datasets is crucial for training deep models, annotating the data, however, inevitably yields noisy labels, which poses challenges to deep learning algorithms. Previous efforts tend to mitigate this problem via identifying and removing noisy samples or correcting their labels according to the statistical properties (e.g., loss values) among training samples. In this paper, we aim to tackle this problem from a new perspective, delving into the deep feature maps, we empirically find that models trained with clean and mislabeled samples manifest distinguishable activation feature distributions. From this observation, a novel robust training approach termed adversarial noisy masking is proposed. The idea is to regularize deep features with a label quality guided masking scheme, which adaptively modulates the input data and label simultaneously, preventing the model to overfit noisy samples. Further, an auxiliary task is designed to reconstruct input data, it naturally provides noise-free self-supervised signals to reinforce the generalization ability of deep models. The proposed method is simple and flexible, it is tested on both synthetic and real-world noisy datasets, where significant improvements are achieved over previous state-of-the-art methods

    Rethinking Mobile Block for Efficient Attention-based Models

    Full text link
    This paper focuses on developing modern, efficient, lightweight models for dense predictions while trading off parameters, FLOPs, and performance. Inverted Residual Block (IRB) serves as the infrastructure for lightweight CNNs, but no counterpart has been recognized by attention-based studies. This work rethinks lightweight infrastructure from efficient IRB and effective components of Transformer from a unified perspective, extending CNN-based IRB to attention-based models and abstracting a one-residual Meta Mobile Block (MMB) for lightweight model design. Following simple but effective design criterion, we deduce a modern Inverted Residual Mobile Block (iRMB) and build a ResNet-like Efficient MOdel (EMO) with only iRMB for down-stream tasks. Extensive experiments on ImageNet-1K, COCO2017, and ADE20K benchmarks demonstrate the superiority of our EMO over state-of-the-art methods, e.g., EMO-1M/2M/5M achieve 71.5, 75.1, and 78.4 Top-1 that surpass equal-order CNN-/Attention-based models, while trading-off the parameter, efficiency, and accuracy well: running 2.8-4.0x faster than EdgeNeXt on iPhone14

    AR-Miner: Mining informative reviews for developers from mobile app marketplace

    Get PDF
    Ministry of Education, Singapore under its Academic Research Funding Tier

    Enabling 3d Printing Technologies For Artificial Compound Eye System And Penetrating Neural Probes

    No full text
    3D printing has become a useful and transformative method and has applications in many different fields, including organ printing, aerospace applications, and medical devices. With higher resolution, faster production speed, and more design flexibility, 3D printing technologies can lead to more novel devices and systems. In this research, two new techniques have been developed to enable 3D printing technologies for artificial compound eye system and penetrating glassy carbon neural electrode array. This work focuses on developing new applications based on the SLA 3D printing process, which uses the liquid resin to create a solid 3D structure. The limitation of SLA 3D printing was addressed by integrating synergistic techniques into the two new approaches. The first one is the application of SLA 3D printing technology together with a polymer filling process for lenses to fabricate 3D structures that mimic an artificial compound eye system. The second work is the application of DLP 3D printing technology together with a pyrolysis process to fabricate the penetrating electrical-conducting carbon neural probes arrays

    Deep learning with application to hashing

    No full text
    Deep Learning and Learning to Hash are two important research areas in machine learning, which have rapid improvements in recent years. What I mainly researched on is an inter-discipline field: deep learning for cross view hashing. Multiple layers of representation in deep learning has the property of abstracting representation from input data, while, in the cross view similarity search, the biggest difficulty is to represent items from one domain to another. Here, I want to take advantage of the latest deep learning technology to solve the cross view similarity search problem. Hashing is used to accelerate this process. This thesis mainly contains three parts. Chapter 2 is a literature survey. It contains a deep learning survey and a learning to hash survey. The deep learning survey briefly introduces fundamental technology of deep learning and its recent development including the latest technology. The Learning to Hash survey brief introduces some widely used learning to hash algorithms. Chapter 3 is an experiment about comparison of some state of the arts learning to hash algorithms. Chapter 4 is cross view hashing based on deep learning. I present a cross view feature hashing technique using deep learning and show some results. These three chapters are main chapters. Chapter 1 and Chapter 5 are introduction and conclusion.MASTER OF ENGINEERING (SCE

    Health Condition Assessment of Marine Systems Based on an Improved Radar Chart

    No full text
    Since health assessment plays a significant role in marine systems (MSs), it has caught the attention of researchers. In this study, a powerful evaluation method called an improved radar chart was developed as a means of reliability estimation. General evaluation methods applied in the comprehensive evaluation of MS are slightly insufficient in terms of considering index coordination. However, the application of a radar chart can solve this problem. To improving the shortcomings of a traditional radar chart, the fuzzy centralization statistical theory and the entropy weight were combined in this study to obtain the comprehensive weight. The weight could be converted into an angle, and it could reflect the influence degree of the indexes on the evaluation objects. Additionally, an angle bisector was introduced as an index axis, and the eigenvector was extracted to get the unique evaluation result. The result showed that the proposed method could achieve the continuous online monitoring of the system state, and the reliable and accurate assessment results were able to provide a reference for condition-based maintenance and decision-making

    PCViT: A Pre-Convolutional ViT Coal Gangue Identification Method

    No full text
    For the study of coal and gangue identification using near-infrared reflection spectroscopy, samples of anthracite coal and gangue with similar appearances were collected, and different dust concentrations (200 ug/m3, 500 ug/m3 and 800 ug/m3), detection distances (1.2 m, 1.5 m and 1.8 m) and mixing gangue rates (one-third coal, two-thirds coal, full coal) were collected in the laboratory by the reflection spectroscopy acquisition device and the gangue reflection spectral data. The spectral data were pre-processed using three methods, first-order differentiation, second-order differentiation and standard normal variable transformation, in order to enhance the absorption characteristics of the reflectance spectra and to eliminate the effects of changes in the experimental environment. The PCViT gangue identification model is established, and the disadvantages of the violent patch embedding of the ViT model are improved by using the stepwise convolution operation to extract features. Then, the interdependence of the features of the hyperspectral data is modeled by the self-attention module, and the learned features are optimized adaptively. The results of gangue recognition under nine working conditions show that the proposed recognition model can significantly improve the recognition accuracy, and this study can provide a reference value for gangue recognition using the near-infrared reflection spectra of gangue

    A novel, effective machine learning-based RNA editing profile for predicting the prognosis of lower-grade gliomas

    No full text
    Patients with low-grade glioma (LGG) may survive for long time periods, but their tumors often progress to higher-grade lesions. Currently, no cure for LGG is available. A-to-I RNA editing accounts for nearly 90% of all RNA editing events in humans and plays a role in tumorigenesis in various cancers. However, little is known regarding its prognostic role in LGG. On the basis of The Cancer Genome Atlas (TCGA) data, we used LASSO and univariate Cox regression to construct an RNA editing site signature. The results derived from the TCGA dataset were further validated with Gene Expression Omnibus (GEO) and Chinese Glioma Genome Atlas (CGGA) datasets. Five machine learning algorithms (Decision Trees C5.0, XGboost, GBDT, Lightgbm, and Catboost) were used to confirm the prognosis associated with the RNA editing site signature. Finally, we explored immune function, immunotherapy, and potential therapeutic agents in the high- and low-risk groups by using multiple biological prediction websites. A total of 22,739 RNA editing sites were identified, and a signature model consisting of four RNA editing sites (PRKCSH|chr19:11561032, DSEL|chr18:65174489, UGGT1|chr2:128952084, and SOD2|chr6:160101723) was established. Cox regression analysis indicated that the RNA editing signature was an independent prognostic factor, according to the ROC curve (AUC = 0.823), and the nomogram model had good predictive power (C-index = 0.824). In addition, the predictive ability of the RNA editing signature was confirmed with the machine learning model. The sensitivity of PCI-34051 and Elephantin was significantly higher in the high-risk group than the low-risk group, thus potentially providing a marker to predict the effects of lung cancer drug treatment. RNA editing may serve as a novel survival prediction tool, thus offering hope for developing editing-based therapeutic strategies to combat LGG progression. In addition, this tool may help optimize survival risk assessment and individualized care for patients with low-grade gliomas
    corecore